243 research outputs found
An ELU Network with Total Variation for Image Denoising
In this paper, we propose a novel convolutional neural network (CNN) for
image denoising, which uses exponential linear unit (ELU) as the activation
function. We investigate the suitability by analyzing ELU's connection with
trainable nonlinear reaction diffusion model (TNRD) and residual denoising. On
the other hand, batch normalization (BN) is indispensable for residual
denoising and convergence purpose. However, direct stacking of BN and ELU
degrades the performance of CNN. To mitigate this issue, we design an
innovative combination of activation layer and normalization layer to exploit
and leverage the ELU network, and discuss the corresponding rationale.
Moreover, inspired by the fact that minimizing total variation (TV) can be
applied to image denoising, we propose a TV regularized L2 loss to evaluate the
training effect during the iterations. Finally, we conduct extensive
experiments, showing that our model outperforms some recent and popular
approaches on Gaussian denoising with specific or randomized noise levels for
both gray and color images.Comment: 10 pages, Accepted by the 24th International Conference on Neural
Information Processing (2017
A Research and Strategy of Remote Sensing Image Denoising Algorithms
Most raw data download from satellites are useless, resulting in transmission
waste, one solution is to process data directly on satellites, then only
transmit the processed results to the ground. Image processing is the main data
processing on satellites, in this paper, we focus on image denoising which is
the basic image processing. There are many high-performance denoising
approaches at present, however, most of them rely on advanced computing
resources or rich images on the ground. Considering the limited computing
resources of satellites and the characteristics of remote sensing images, we do
some research on these high-performance ground image denoising approaches and
compare them in simulation experiments to analyze whether they are suitable for
satellites. According to the analysis results, we propose two feasible image
denoising strategies for satellites based on satellite TianZhi-1.Comment: 9 pages, 4 figures, ICNC-FSKD 201
Revisiting loss-specific training of filter-based MRFs for image restoration
It is now well known that Markov random fields (MRFs) are particularly
effective for modeling image priors in low-level vision. Recent years have seen
the emergence of two main approaches for learning the parameters in MRFs: (1)
probabilistic learning using sampling-based algorithms and (2) loss-specific
training based on MAP estimate. After investigating existing training
approaches, it turns out that the performance of the loss-specific training has
been significantly underestimated in existing work. In this paper, we revisit
this approach and use techniques from bi-level optimization to solve it. We
show that we can get a substantial gain in the final performance by solving the
lower-level problem in the bi-level framework with high accuracy using our
newly proposed algorithm. As a result, our trained model is on par with highly
specialized image denoising algorithms and clearly outperforms
probabilistically trained MRF models. Our findings suggest that for the
loss-specific training scheme, solving the lower-level problem with higher
accuracy is beneficial. Our trained model comes along with the additional
advantage, that inference is extremely efficient. Our GPU-based implementation
takes less than 1s to produce state-of-the-art performance.Comment: 10 pages, 2 figures, appear at 35th German Conference, GCPR 2013,
Saarbr\"ucken, Germany, September 3-6, 2013. Proceeding
QuaSI: Quantile Sparse Image Prior for Spatio-Temporal Denoising of Retinal OCT Data
Optical coherence tomography (OCT) enables high-resolution and non-invasive
3D imaging of the human retina but is inherently impaired by speckle noise.
This paper introduces a spatio-temporal denoising algorithm for OCT data on a
B-scan level using a novel quantile sparse image (QuaSI) prior. To remove
speckle noise while preserving image structures of diagnostic relevance, we
implement our QuaSI prior via median filter regularization coupled with a Huber
data fidelity model in a variational approach. For efficient energy
minimization, we develop an alternating direction method of multipliers (ADMM)
scheme using a linearization of median filtering. Our spatio-temporal method
can handle both, denoising of single B-scans and temporally consecutive
B-scans, to gain volumetric OCT data with enhanced signal-to-noise ratio. Our
algorithm based on 4 B-scans only achieved comparable performance to averaging
13 B-scans and outperformed other current denoising methods.Comment: submitted to MICCAI'1
Phase-based video motion processing
We introduce a technique to manipulate small movements in videos based on an analysis of motion in complex-valued image pyramids. Phase variations of the coefficients of a complex-valued steerable pyramid over time correspond to motion, and can be temporally processed and amplified to reveal imperceptible motions, or attenuated to remove distracting changes. This processing does not involve the computation of optical flow, and in comparison to the previous Eulerian Video Magnification method it supports larger amplification factors and is significantly less sensitive to noise. These improved capabilities broaden the set of applications for motion processing in videos. We demonstrate the advantages of this approach on synthetic and natural video sequences, and explore applications in scientific analysis, visualization and video enhancement.Shell ResearchUnited States. Defense Advanced Research Projects Agency. Soldier Centric Imaging via Computational CamerasNational Science Foundation (U.S.) (CGV-1111415)Cognex CorporationMicrosoft Research (PhD Fellowship)American Society for Engineering Education. National Defense Science and Engineering Graduate Fellowshi
Provenance analysis for instagram photos
As a feasible device fingerprint, sensor pattern noise (SPN) has been proven to be effective in the provenance analysis of digital images. However, with the rise of social media, millions of images are being uploaded to and shared through social media sites every day. An image downloaded from social networks may have gone through a series of unknown image manipulations. Consequently, the trustworthiness of SPN has been challenged in the provenance analysis of the images downloaded from social media platforms. In this paper, we intend to investigate the effects of the pre-defined Instagram images filters on the SPN-based image provenance analysis. We identify two groups of filters that affect the SPN in quite different ways, with Group I consisting of the filters that severely attenuate the SPN and Group II consisting of the filters that well preserve the SPN in the images. We further propose a CNN-based classifier to perform filter-oriented image categorization, aiming to exclude the images manipulated by the filters in Group I and thus improve the reliability of the SPN-based provenance analysis. The results on about 20, 000 images and 18 filters are very promising, with an accuracy higher than 96% in differentiating the filters in Group I and Group II
Deep Burst Denoising
Noise is an inherent issue of low-light image capture, one which is
exacerbated on mobile devices due to their narrow apertures and small sensors.
One strategy for mitigating noise in a low-light situation is to increase the
shutter time of the camera, thus allowing each photosite to integrate more
light and decrease noise variance. However, there are two downsides of long
exposures: (a) bright regions can exceed the sensor range, and (b) camera and
scene motion will result in blurred images. Another way of gathering more light
is to capture multiple short (thus noisy) frames in a "burst" and intelligently
integrate the content, thus avoiding the above downsides. In this paper, we use
the burst-capture strategy and implement the intelligent integration via a
recurrent fully convolutional deep neural net (CNN). We build our novel,
multiframe architecture to be a simple addition to any single frame denoising
model, and design to handle an arbitrary number of noisy input frames. We show
that it achieves state of the art denoising results on our burst dataset,
improving on the best published multi-frame techniques, such as VBM4D and
FlexISP. Finally, we explore other applications of image enhancement by
integrating content from multiple frames and demonstrate that our DNN
architecture generalizes well to image super-resolution
A GPU-accelerated real-time NLMeans algorithm for denoising color video sequences
Abstract. The NLMeans filter, originally proposed by Buades et al., is a very popular filter for the removal of white Gaussian noise, due to its simplicity and excellent performance. The strength of this filter lies in exploiting the repetitive character of structures in images. However, to fully take advantage of the repetitivity a computationally extensive search for similar candidate blocks is indispensable. In previous work, we presented a number of algorithmic acceleration techniques for the NLMeans filter for still grayscale images. In this paper, we go one step further and incorporate both temporal information and color information into the NLMeans algorithm, in order to restore video sequences. Starting from our algorithmic acceleration techniques, we investigate how the NLMeans algorithm can be easily mapped onto recent parallel computing architectures. In particular, we consider the graphical processing unit (GPU), which is available on most recent computers. Our developments lead to a high-quality denoising filter that can process DVD-resolution video sequences in real-time on a mid-range GPU
Space-Variant Gabor Decomposition for Filtering 3D Medical Images
This is an experimental paper in which we introduce the possibility to analyze and to synthesize 3D medical images by using multivariate Gabor frames with Gaussian windows. Our purpose is to apply a space-variant filter-like operation in the space-frequency domain to correct medical images corrupted by different types of acquisitions errors. The Gabor frames are constructed with Gaussian windows sampled on non-separable lattices for a better packing of the space-frequency plane. An implementable solution for 3D-Gabor frames with non-separable lattice is given and numerical tests on simulated data are presented.Austrian Science Fund (FWF) P2751
GIA-Net: Global Information Aware Network for Low-light Imaging
It is extremely challenging to acquire perceptually plausible images under
low-light conditions due to low SNR. Most recently, U-Nets have shown promising
results for low-light imaging. However, vanilla U-Nets generate images with
artifacts such as color inconsistency due to the lack of global color
information. In this paper, we propose a global information aware (GIA) module,
which is capable of extracting and integrating the global information into the
network to improve the performance of low-light imaging. The GIA module can be
inserted into a vanilla U-Net with negligible extra learnable parameters or
computational cost. Moreover, a GIA-Net is constructed, trained and evaluated
on a large scale real-world low-light imaging dataset. Experimental results
show that the proposed GIA-Net outperforms the state-of-the-art methods in
terms of four metrics, including deep metrics that measure perceptual
similarities. Extensive ablation studies have been conducted to verify the
effectiveness of the proposed GIA-Net for low-light imaging by utilizing global
information.Comment: 16 pages 6 figures; accepted to AIM at ECCV 202
- …